SelfCCL: Curriculum Contrastive Learning by Transferring Self-Taught Knowledge for Fine-Tuning BERT
نویسندگان
چکیده
BERT, the most popular deep learning language model, has yielded breakthrough results in various NLP tasks. However, semantic representation space learned by BERT property of anisotropy. Therefore, needs to be fine-tuned for certain downstream tasks such as Semantic Textual Similarity (STS). To overcome this problem and improve sentence space, some contrastive methods have been proposed fine-tuning BERT. existing models do not consider importance input triplets terms easy hard negatives during training. In paper, we propose SelfCCL: Curriculum Contrastive Learning model Transferring Self-taught Knowledge Fine-Tuning which mimics two ways that humans learn about world around them, namely curriculum learning. The former learns contrasting similar dissimilar samples. latter is inspired way from simplest concepts complex concepts. Our also performs training transferring self-taught knowledge. That is, figures out are or difficult based on previously knowledge, then those order using a objective. We apply our Sentence BERT(SBERT) frameworks. evaluation SelfCCL standard STS SentEval transfer show together with increases average performance extent.
منابع مشابه
Risk Bounds for Transferring Representations With and Without Fine-Tuning
A popular machine learning strategy is the transfer of a representation (i.e. a feature extraction function) learned on a source task to a target task. Examples include the re-use of neural network weights or word embeddings. We develop sufficient conditions for the success of this approach. If the representation learned from the source task is fixed, we identify conditions on how the tasks rel...
متن کاملAn Interactive Knowledge-based Learning Environment for Transferring Health Knowledge
This paper describes a narrative based interactive intelligent learning environment which aims to elucidate reasoning using interactive scenarios that may be used in training novices ICU nurses in the application of their knowledge and decision making. Its design is based on an approach to generating narrative from knowledge that has been modelled in specific decision/reasoning domains. The app...
متن کاملRobust and Discriminative Self-Taught Learning
The lack of training data is a common challenge in many machine learning problems, which is often tackled by semi-supervised learning methods or transfer learning methods. The former requires unlabeled images from the same distribution as the labeled ones and the latter leverages labeled images from related homogenous tasks. However, these restrictions often cannot be satisfied. To address this...
متن کاملSelf-Paced Curriculum Learning
Curriculum learning (CL) or self-paced learning (SPL) represents a recently proposed learning regime inspired by the learning process of humans and animals that gradually proceeds from easy to more complex samples in training. The two methods share a similar conceptual learning paradigm, but differ in specific learning schemes. In CL, the curriculum is predetermined by prior knowledge, and rema...
متن کاملDeep Self-Taught Learning for Handwritten Character Recognition
Recent theoretical and empirical work in statistical machine learning has demonstrated the importance of learning algorithms for deep architectures, i.e., function classes obtained by composing multiple non-linear transformations. Self-taught learning (exploiting unlabeled examples or examples from other distributions) has already been applied to deep learners, but mostly to show the advantage ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Applied sciences
سال: 2023
ISSN: ['2076-3417']
DOI: https://doi.org/10.3390/app13031913